Compressed Least-Squares Regression
نویسندگان
چکیده
We consider the problem of learning, from K data, a regression function in a linear space of high dimensionN using projections onto a random subspace of lower dimension M . From any algorithm minimizing the (possibly penalized) empirical risk, we provide bounds on the excess risk of the estimate computed in the projected subspace (compressed domain) in terms of the excess risk of the estimate built in the high-dimensional space (initial domain). We show that solving the problem in the compressed domain instead of the initial domain reduces the estimation error at the price of an increased (but controlled) approximation error. We apply the analysis to Least-Squares (LS) regression and discuss the excess risk and numerical complexity of the resulting “Compressed Least Squares Regression” (CLSR) in terms of N , K, and M . When we choose M = O( √ K), we show that CLSR has an estimation error of order O(logK/ √ K).
منابع مشابه
Compressed Least Squares Regression revisited
We revisit compressed least squares (CLS) regression as originally analyzed in Maillard and Munos (2009) and later on in Kaban (2014) with some refinements. Given a set of high-dimensional inputs, CLS applies a random projection and then performs least squares regression based on the projected inputs of lower dimension. This approach can be beneficial with regard to both computation (yielding a...
متن کاملar X iv : 0 70 6 . 05 34 v 1 [ st at . M L ] 4 J un 2 00 7 Compressed Regression
Recent research has studied the role of sparsity in high dimensional regression and signal reconstruction, establishing theoretical limits for recovering sparse models from sparse data. This line of work shows that l1-regularized least squares regression can accurately estimate a sparse linear model from n noisy examples in p dimensions, even if p is much larger than n. In this paper we study a...
متن کاملAnnotated Bibliography High-dimensional Statistical Inference
Recent research has studied the role of sparsity in high dimensional regression and signal reconstruction, establishing theoretical limits for recovering sparse models from sparse data. This line of work shows that l1-regularized least squares regression can accurately estimate a sparse linear model from n noisy examples in p dimensions, even if p is much larger than n. In this paper we study a...
متن کاملAn Enhanced Dimension Reduction Approach for Microarray Gene Expression Data
Methods: Here, we introduce an enhancement concept that joins two dimension reduction methods of Partial Least Squares (PLS) and Minimum Average Variance Estimation (MAVE), which is called as an enhanced dimension reduction method. The PLS method generates the new transformed genes that include compressed information. Then, the MAVE method clusters samples in the same class and separate samples...
متن کاملTitle Placeholder
This paper compares and discusses four techniques for model order reduction based on compressed sensing (CS), less relevant basis removal (LRBR), principal component analysis (PCA) and partial least squares (PLS). CS and PCA have already been used for reducing the order of power amplifier (PA) behavioral models for digital predistortion (DPD) purposes. While PLS, despite being popular in some s...
متن کامل